perm filename CHAP4A[4,KMC] blob
sn#016893 filedate 1972-12-19 generic text, type T, neo UTF8
00100 PROBLEMS OF NATURAL LANGUAGE UNDERSTANDING IN TELETYPED INTERVIEW DIALOGUES.
00200
00300
00400 By `natural language` I shall mean everyday American English
00500 such as is used by readers of this book in ordinary conversations.
00600 It is still difficult to be explicit about the processes which
00700 enable hummans to interpret and respond to natural language.
00800 Philosophers, linguists and psychologists have speculated about
00900 and investigated natural language with various purposes and few
01000 useful results. Now attempts are being made in artificial intelligence to write
01100 algorithims which `understand' what is being expressed in natural
01200 language utterances.
01300 During the 1960's when machine processing of natural language
01310
01400 was dominated by syntactic considerations, it became clear that
01500 this approach was insufficient. The current view is that to understand
01600 what utterances say, knowledGe about linguistic syntax and semantics
01700 must be combined with knowledge from an underlying conceptual
01800 structure containing a world-model and an ability to draw inferences.
01900 How to achieve this combination efficiently represents a huge task for
02000 both theory and implementation.
02100 Since the behavior being simulated by our paranoid model is the
02200 linguistic(non-nonverbal) behavior of paranoid patients in a psychiatric
02300 interview, the model must have some ability to process and respond to
02400 natural language input in a manner indicating the underlying pathological
02500 beliefs characteristic of the paranoid mode. This our purpose was
02600 to develop a method for understanding everyday English sufficient
02700 for the model to behave conversationally in a paranoid way in a
02800 circumscribed situation. What is said in this situation is far
02900 richer than what is said in conversations with a block-stacking
03000
03100 robot but its requirements for constructing an interpretation
03200 of an input are not as complex as trying to understand anything
03300 said in English BY anybody in any dialogue situation.
03400 We took a pragmatic approach which considered "understanding"
03500 to represent "getting the message" of an utterance by
03600 gleaning some {not all} of the relations between them.
03700 this straightforward approach to a complex problem has its
03800 drawbacks, as will be shown, but we were striving for a
03900 sufficiency to demonstrate paranoia rather than complete
04000 comprehension of English.
04100 Linguistic approaches cite traditional problems with
04200 ambiguity, as illustrated in the following example from
04300 Wilks { }. Suppose I walked up to you, a stranger, on
04400 the street on Sunday morning and said
04500 {1} He fell while getting to the ball.'
04600 Admittedly this is a strange scene and in this situation
04700 you would think me to be crazy, hungover and maybe still
04800 drunk, but the example is no more weird than the isolated
04900 examples discussed in the linguistics literature. Suppose
05000 further that in your personal `dictionary' the word 'ball'
05100 has at least two senses, {A} a spherical physical object
05200 used in a game, and {B} a formal dance. {It probably has
05300 also a third sense as a verb but we will ignore this more
05400 or less recent example of semantic shift}. Having no
05500 further information in this situation and attempting to
05600 construct an interpretation of my utterance, you would be
05700 puzzled as to whether I was referring to a ball game or a
05800 dance. If we then continued on our respective ways, saying
05900 nothing else, your puzzlement would continue and even increase
06000 --I don't know what he was referring to nor why he even said
06100 that to me.
06200
06300 The ambiguity arises because of the two word senses for
06400 ball, each of which would give the utterance a meaningful
06500 interpretation. But the example is extremely forced and
06600 artificial. Such isolated utterances cannot be disambiguated
06700 {uniqueated is a better term} but this is no handicap for
06800 ordinary human convversations in which ambiguities hardly arise
06900 at all. Besides the utterance itself, extra information is
07000 usually available in the form of contextual and situational
07100 knowledge. Even better, one can always ask. If I had said
07200 only utterance {1} to you, you could simply ask:
07300
07400 {2} `What do you mean?'
07500
07600 and my reply would indicate something about a game or a
07700 dance or who `he' was.
07800 Utterances occur in conversations which take place in
07900 sociopsychological situations. The communicants have roles
08000 and intentions towards one another. If the situation is that
08100 of a medical or psychiatric interview between doctor and
08200 patient and the doctor asks:
08300
08400 {3} `How much do you drink?'
08410
08500 We know from the nature of the situation that drink means
08600 `drink alcohol' and does not refer to a total fluid intake.
08700
08800 Dialogues represent connected discourse in which all
08900 the utterances, except perhaps for opening greetings, are
09000 connectable to previous utterances, contexts and sub
09100 contexts, topics and subtopics surround any given utterance
09200 and activate relevant word senses such that alternative
09300 senses do not arise in the comprehension process. In
09400 spoken dialogues intonations and word emphases are further
09500 means for avoiding ambbiguities, but the connected discourse
09600 of dialogues brings problems of its own to the algorithmist
09700 whose program must keep track of what is going on and what
09800 as been said before. Foremost is the problem of anaphoric
09900 reference.
10000
10050 ANAPHORA
10100 An anaphoric reference is a word or phrase which refers
10200 backwards {usually--i.e. there are some rare forward references}
10300 to something in a previous utterance. A common example in
10400 intervview dialogues is that of pronouns.
10500
10600 {4} Patient - `My father was an alcoholic.'
10700
10800 {5} Doctor - `Were you very close to him?'
10900
11000 Where the term `him' in {4} obviously refers to `father' in
11100 {3}, it is not too difficult a probblem for a program to make
11200 the correct assignments in personal pronoun anaphora. Of
11300 greater compexity are utterances containing the words 'it'
11400 or `this'. For example, suppose the interview continued after
11500 {5} as follows:
11600
11700 {6} Patient - `Yes I was, even though his drinking
11800 upset me.'
11900
12000 {7} Doctor - `How did it upset you?'
12100
12200 Here the comprehension algorithm must grasp that the `it' of
12300 {7} refers to the topic of the father's drinking in {6}.
12400 Further, if the dialogue continued:
12500
12600 {8}Patient - 'It embarassed me when my friends saw
12700 him drunk.
12800
12900 {9} Doctor - `Do you think he sensed this.'
13000
13100 then it must bbe understood that `this' refers to the patient's
13200 embarassment. How this is implemented in the program will be
13300 described in xxx.
13400
13500 FRAGMENTS
13600
13700 Another major probblem for algorithms which attempt to understand
13800 discourse consists of the fact that many of the input expressions
13900 are not well-formed. All sorts of fragments and ellipses appear
14000 which must somehow be connected to conceptualizations under discussion.
14100 For example, consider the following exchange:
14200
14300
14400 {10} Dr. - `How do you like the hospital?'
14500
14600 {11} Pt. - `I shouldn't be here.'
14700
14800 {12} Dr. - `Why not?'
14900 The question {12} is an elliptical expression for the full concept
15000 ualization
15100
15200 `Why should you not be in the hospital?'
15300
15400 Junk words {`well now'} {`tell me more'} and go ahead signals
15500 must be responded to by continuation of a topic.
15600
15700 For example:
15800
15900 {13} Pt.- `I went to the track last week.'
16000
16100 {14} Dr. - `Really?'
16200
16300 Such expressions as {14} stand in a meta-relation to the topic and
16400 serve to keep the conversation going.
16500
16600 REJOINDERS
16700
16800 Sometimes the input expression from the interviewer is a rejoinder
16900 , a reply to a reply by the patient. For instance:
17000
17100 {15} Dr. - `Who are you afraid of?'
17200
17300 {16} Pt. - `The Mafia is out to get me.'
17400
17500 {17} Dr. - `I would bbe afraid of them also.'
17600
17700 in which {17} is a rejoinder. Such expressions are not requests for
17800 information but provide information for the patient's model of the
17900 interviewer.
18000
18100 INTERVIEWER-INTERVIEWEE RELATIONS
18200
18300 It is characteristic of psychiatric interviewing that the
18400 participants from time to time do not simply talk about the
18500 patient. Two situations exist concurrently in an interview,
18600 one being talked about and one the participants are in. At
18700 times the second situation becomes the first. When the partici
18800 pants discuss one another and their relation, the dialogue
18900 expressions contain intentional verbs which in English fit the
19000 pattern `I X you' or `you X me'. The comprehension process must
19100 distinguish clearly between subbject and object in the case of some
19200 of these verbs. For example in
19300
19400 {18} `I like you'
19500
19600 the speaker 'I' experiences the liking but in
19700
19800 {19} `Do I please you?'
19900
20000 the `you' experiences the pleasure as a consequence of something
20100 `I' does.